4 research outputs found
Light field super resolution through controlled micro-shifts of light field sensor
Light field cameras enable new capabilities, such as post-capture refocusing
and aperture control, through capturing directional and spatial distribution of
light rays in space. Micro-lens array based light field camera design is often
preferred due to its light transmission efficiency, cost-effectiveness and
compactness. One drawback of the micro-lens array based light field cameras is
low spatial resolution due to the fact that a single sensor is shared to
capture both spatial and angular information. To address the low spatial
resolution issue, we present a light field imaging approach, where multiple
light fields are captured and fused to improve the spatial resolution. For each
capture, the light field sensor is shifted by a pre-determined fraction of a
micro-lens size using an XY translation stage for optimal performance
Light-field view synthesis using convolutional block attention module
Consumer light-field (LF) cameras suffer from a low or limited resolution
because of the angular-spatial trade-off. To alleviate this drawback, we
propose a novel learning-based approach utilizing attention mechanism to
synthesize novel views of a light-field image using a sparse set of input views
(i.e., 4 corner views) from a camera array. In the proposed method, we divide
the process into three stages, stereo-feature extraction, disparity estimation,
and final image refinement. We use three sequential convolutional neural
networks for each stage. A residual convolutional block attention module (CBAM)
is employed for final adaptive image refinement. Attention modules are helpful
in learning and focusing more on the important features of the image and are
thus sequentially applied in the channel and spatial dimensions. Experimental
results show the robustness of the proposed method. Our proposed network
outperforms the state-of-the-art learning-based light-field view synthesis
methods on two challenging real-world datasets by 0.5 dB on average.
Furthermore, we provide an ablation study to substantiate our findings